24 research outputs found

    Joint 3D estimation of vehicles and scene flow

    Get PDF
    driving. While much progress has been made in recent years, imaging conditions in natural outdoor environments are still very challenging for current reconstruction and recognition methods. In this paper, we propose a novel unified approach which reasons jointly about 3D scene flow as well as the pose, shape and motion of vehicles in the scene. Towards this goal, we incorporate a deformable CAD model into a slanted-plane conditional random field for scene flow estimation and enforce shape consistency between the rendered 3D models and the parameters of all superpixels in the image. The association of superpixels to objects is established by an index variable which implicitly enables model selection. We evaluate our approach on the challenging KITTI scene flow dataset in terms of object and scene flow estimation. Our results provide a prove of concept and demonstrate the usefulness of our method. © 2015 Copernicus GmbH. All Rights Reserved

    An iterative inference procedure applying conditional random fields for simultaneous classification of land cover and land use

    Get PDF
    Land cover and land use exhibit strong contextual dependencies. We propose a novel approach for the simultaneous classification of land cover and land use, where semantic and spatial context is considered. The image sites for land cover and land use classification form a hierarchy consisting of two layers: a land cover layer and a land use layer. We apply Conditional Random Fields (CRF) at both layers. The layers differ with respect to the image entities corresponding to the nodes, the employed features and the classes to be distinguished. In the land cover layer, the nodes represent super-pixels; in the land use layer, the nodes correspond to objects from a geospatial database. Both CRFs model spatial dependencies between neighbouring image sites. The complex semantic relations between land cover and land use are integrated in the classification process by using contextual features. We propose a new iterative inference procedure for the simultaneous classification of land cover and land use, in which the two classification tasks mutually influence each other. This helps to improve the classification accuracy for certain classes. The main idea of this approach is that semantic context helps to refine the class predictions, which, in turn, leads to more expressive context information. Thus, potentially wrong decisions can be reversed at later stages. The approach is designed for input data based on aerial images. Experiments are carried out on a test site to evaluate the performance of the proposed method. We show the effectiveness of the iterative inference procedure and demonstrate that a smaller size of the super-pixels has a positive influence on the classification result

    SATSal: A Multi-Level Self-Attention Based Architecture for Visual Saliency Prediction

    Get PDF
    Human visual Attention modelling is a persistent interdisciplinary research challenge, gaining new interest in recent years mainly due to the latest developments in deep learning. That is particularly evident in saliency benchmarks. Novel deep learning-based visual saliency models show promising results in capturing high-level (top-down) human visual attention processes. Therefore, they strongly differ from the earlier approaches, mainly characterised by low-level (bottom-up) visual features. These developments account for innate human selectivity mechanisms that are reliant on both high- and low-level factors. Moreover, the two factors interact with each other. Motivated by the importance of these interactions, in this project, we tackle visual saliency modelling holistically, examining if we could consider both high- and low-level features that govern human attention. Specifically, we propose a novel method SAtSal (Self-Attention Saliency). SAtSal leverages both high and low-level features using a multilevel merging of skip connections during the decoding stage. Consequently, we incorporate convolutional self-attention modules on skip connection from the encoder to the decoder network to properly integrate the valuable signals from multilevel spatial features. Thus, the self-attention modules learn to filter out the latent representation of the salient regions from the other irrelevant information in an embedded and joint manner with the main encoder-decoder model backbone. Finally, we evaluate SAtSal against various existing solutions to validate our approach, using the well-known standard saliency benchmark MIT300. To further examine SAtSal's robustness on other image types, we also evaluate it on the Le-Meur saliency painting benchmark

    Gaussian process for activity modeling and anomaly detection

    Get PDF
    Complex activity modeling and identification of anomaly is one of the most interesting and desired capabilities for automated video behavior analysis. A number of different approaches have been proposed in the past to tackle this problem. There are two main challenges for activity modeling and anomaly detection: 1) most existing approaches require sufficient data and supervision for learning; 2) the most interesting abnormal activities arise rarely and are ambiguous among typical activities, i.e. hard to be precisely defined. In this paper, we propose a novel approach to model complex activities and detect anomalies by using non-parametric Gaussian Process (GP) models in a crowded and complicated traffic scene. In comparison with parametric models such as HMM, GP models are nonparametric and have their advantages. Our GP models exploit implicit spatial-temporal dependence among local activity patterns. The learned GP regression models give a probabilistic prediction of regional activities at next time interval based on observations at present. An anomaly will be detected by comparing the actual observations with the prediction at real time. We verify the effectiveness and robustness of the proposed model on the QMUL Junction Dataset. Furthermore, we provide a publicly available manually labeled ground truth of this data set

    Towards SDG 4 : trade -offs for geospatial open educational resources

    Get PDF
    Through the ISPRS scientific initiative presented in this paper, we aim to make geospatial educational resources available and discoverable to those who teach and those who want to learn. In earlier work, we designed and implemented a prototype catalogue for geospatial educational resources, aimed at a target audience in higher education. The success of search and discovery in any catalogue relies heavily on the metadata that describes catalogue entries. Initial feedback showed that users find it difficult to use some of the metadata elements in the prototype to describe their teaching materials. In order to better understand their difficulties and further refine the metadata for describing educational resources that are used for geospatial purposes specifically, we asked a number of participants to describe geospatial educational resources according to four sets of metadata attributes. This paper presents the results of the study and recommends a set of metadata attributes that are specifically useful for geospatial educational resources. Implementation trade-offs are discussed, e.g., deciding between metadata attributes that are very specific or more generic, and catalogue entries that are immediately available to Web search engines without any quality checks vs. catalogue entries that are moderated by a community of educators before publishing them. By providing metadata about geospatial educational resources, the international geospatial community can contribute to the UN Sustainable Development Goal 4 to ensure inclusive and equitable quality education and to promote lifelong learning opportunities for all.This research was supported by a grant from the 2018 ISPRS Education and Capacity Building Initiatives. We would also like to thank Azile Mdleleni and Cameron Green who described geospatial educational resources according to the metadata schemas.https://www.isprs.org/publications/archives.aspxpm2021Geography, Geoinformatics and Meteorolog

    Global and local sparse subspace optimization for motion segmentation

    Get PDF
    In this paper, we propose a new framework for segmenting feature-based moving objects under affine subspace model. Since the feature trajectories in practice are high-dimensional and contain a lot of noise, we firstly apply the sparse PCA to represent the original trajectories with a low-dimensional global subspace, which consists of the orthogonal sparse principal vectors. Subsequently, the local subspace separation will be achieved via automatically searching the sparse representation of the nearest neighbors for each projected data. In order to refine the local subspace estimation result, we propose an error estimation to encourage the projected data that span a same local subspace to be clustered together. In the end, the segmentation of different motions is achieved through the spectral clustering on an affinity matrix, which is constructed with both the error estimation and sparse neighbors optimization. We test our method extensively and compare it with state-of-the-art methods on the Hopkins 155 dataset. The results show that our method is comparable with the other motion segmentation methods, and in many cases exceed them in terms of precision and computation time

    Impact analysis of accidents on the traffic flow based on massive floating car data

    Get PDF
    The wide usage of GPS-equipped devices enables the mass recording of vehicle movement trajectories describing the movement behavior of the traffic participants. An important aspect of the road traffic is the impact of anomalies, like accidents, on traffic flow. Accidents are especially important as they contribute to the the aspects of safety and also influence travel time estimations. In this paper, the impact of accidents is determined based on a massive GPS trajectory and accident dataset. Due to the missing precise date of the accidents in the data set used, first, the date of the accident is estimated based on the speed profile at the accident time. Further, the temporal impact of the accident is estimated using the speed profile of the whole day. The approach is applied in an experiment on a one month subset of the datasets. The results show that more than 72% of the accident dates are identified and the impact on the temporal dimension is approximated. Moreover, it can be seen that accidents during the rush hours and on high frequency road types (e.g. motorways, trunks or primaries) have an increasing effect on the impact duration on the traffic flow

    Visual exploration of eye movement data using the space-time-cube

    No full text
    corecore